Bayesian dropout
نویسندگان
چکیده
Dropout has recently emerged as a powerful and simple method for training neural networks preventing co-adaptation by stochastically omitting neurons. is currently not grounded in explicit modelling assumptions which so far precluded its adoption Bayesian modelling. Using entropic reasoning we show that dropout can be interpreted optimal inference under constraints. We demonstrate this on an analytically tractable regression model providing interpretation of mechanism regularizing well connection to other techniques. also discuss two general approximate techniques applying models, one based analytical approximation the stochastic variational These are then applied Baysian logistic problem shown improve performance become more misspecified. Our framework roots theoretically justified practical tool statistical allowing Bayesians tap into benefits training.
منابع مشابه
A Bayesian encourages dropout
Dropout is one of the key techniques to prevent the learning from overfitting. It is explained that dropout works as a kind of modified L2 regularization. Here, we shed light on the dropout from Bayesian standpoint. Bayesian interpretation enables us to optimize the dropout rate, which is beneficial for learning of weight parameters and prediction after learning. The experiment result also enco...
متن کاملDropout as a Bayesian Approximation: Appendix
We show that a neural network with arbitrary depth and non-linearities, with dropout applied before every weight layer, is mathematically equivalent to an approximation to a well known Bayesian model. This interpretation might offer an explanation to some of dropout’s key properties, such as its robustness to overfitting. Our interpretation allows us to reason about uncertainty in deep learning...
متن کاملHigh Dimensional Bayesian Optimization using Dropout
Scaling Bayesian optimization to high dimensions is challenging task as the global optimization of high-dimensional acquisition function can be expensive and often infeasible. Existing methods depend either on limited “active” variables or the additive form of the objective function. We propose a new method for high-dimensional Bayesian optimization, that uses a dropout strategy to optimize onl...
متن کاملDropout Inference in Bayesian Neural Networks with Alpha-divergences
To obtain uncertainty estimates with real-world Bayesian deep learning models, practical inference approximations are needed. Dropout variational inference (VI) for example has been used for machine vision and medical applications, but VI can severely underestimates model uncertainty. Alpha-divergences are alternative divergences to VI’s KL objective, which are able to avoid VI’s uncertainty un...
متن کاملA PAC-Bayesian Tutorial with A Dropout Bound
This tutorial gives a concise overview of existing PAC-Bayesian theory focusing on three generalization bounds. The first is an Occam bound which handles rules with finite precision parameters and which states that generalization loss is near training loss when the number of bits needed to write the rule is small compared to the sample size. The second is a PAC-Bayesian bound providing a genera...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Procedia Computer Science
سال: 2022
ISSN: ['1877-0509']
DOI: https://doi.org/10.1016/j.procs.2022.03.105